navigation map
NavMapFusion: Diffusion-based Fusion of Navigation Maps for Online Vectorized HD Map Construction
Monninger, Thomas, Zhang, Zihan, Staab, Steffen, Ding, Sihao
Accurate environmental representations are essential for autonomous driving, providing the foundation for safe and efficient navigation. Traditionally, high-definition (HD) maps are providing this representation of the static road infrastructure to the autonomous system a priori. However, because the real world is constantly changing, such maps must be constructed online from on-board sensor data. Navigation-grade standard-definition (SD) maps are widely available, but their resolution is insufficient for direct deployment. Instead, they can be used as coarse prior to guide the online map construction process. We propose NavMapFusion, a diffusion-based framework that performs iterative denoising conditioned on high-fidelity sensor data and on low-fidelity navigation maps. This paper strives to answer: (1) How can coarse, potentially outdated navigation maps guide online map construction? (2) What advantages do diffusion models offer for map fusion? We demonstrate that diffusion-based map construction provides a robust framework for map fusion. Our key insight is that discrepancies between the prior map and online perception naturally correspond to noise within the diffusion process; consistent regions reinforce the map construction, whereas outdated segments are suppressed. On the nuScenes benchmark, NavMapFusion conditioned on coarse road lines from OpenStreetMap data reaches a 21.4% relative improvement on 100 m, and even stronger improvements on larger perception ranges, while maintaining real-time capabilities. By fusing low-fidelity priors with high-fidelity sensor data, the proposed method generates accurate and up-to-date environment representations, guiding towards safer and more reliable autonomous driving. The code is available at https://github.com/tmonnin/navmapfusion
- North America > United States > California > San Diego County > San Diego (0.04)
- Europe > United Kingdom > England > Hampshire > Southampton (0.04)
- Europe > Germany > Baden-Württemberg > Stuttgart Region > Stuttgart (0.04)
- Asia > Middle East > Jordan (0.04)
- Transportation > Ground > Road (0.69)
- Information Technology (0.55)
- Automobiles & Trucks (0.55)
EfficientNav: Towards On-Device Object-Goal Navigation with Navigation Map Caching and Retrieval
Yang, Zebin, Zheng, Sunjian, Xie, Tong, Xu, Tianshi, Yu, Bo, Wang, Fan, Tang, Jie, Liu, Shaoshan, Li, Meng
Object-goal navigation (ObjNav) tasks an agent with navigating to the location of a specific object in an unseen environment. Embodied agents equipped with large language models (LLMs) and online constructed navigation maps can perform ObjNav in a zero-shot manner. However, existing agents heavily rely on giant LLMs on the cloud, e.g., GPT-4, while directly switching to small LLMs, e.g., LLaMA3.2-11b, suffer from significant success rate drops due to limited model capacity for understanding complex navigation maps, which prevents deploying ObjNav on local devices. At the same time, the long prompt introduced by the navigation map description will cause high planning latency on local devices. In this paper, we propose EfficientNav to enable on-device efficient LLM-based zero-shot ObjNav. To help the smaller LLMs better understand the environment, we propose semantics-aware memory retrieval to prune redundant information in navigation maps. To reduce planning latency, we propose discrete memory caching and attention-based memory clustering to efficiently save and re-use the KV cache. Extensive experimental results demonstrate that EfficientNav achieves 11.1% improvement in success rate on HM3D benchmark over GPT-4-based baselines, and demonstrates 6.7x real-time latency reduction and 4.7x end-to-end latency reduction over GPT-4 planner. Our code is available on https://github.com/PKU-SEC-Lab/EfficientNav.
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Asia > China > Beijing > Beijing (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- (2 more...)
Stereovision Image Processing for Planetary Navigation Maps with Semi-Global Matching and Superpixel Segmentation
Lu, Yan-Shan, Arana-Catania, Miguel, Upadhyay, Saurabh, Felicetti, Leonard
Mars exploration requires precise and reliable terrain models to ensure safe rover navigation across its unpredictable and often hazardous landscapes. Stereoscopic vision serves a critical role in the rover's perception, allowing scene reconstruction by generating precise depth maps through stereo matching. State-of-the-art Martian planetary exploration uses traditional local block-matching, aggregates cost over square windows, and refines disparities via smoothness constraints. However, this method often struggles with low-texture images, occlusion, and repetitive patterns because it considers only limited neighbouring pixels and lacks a wider understanding of scene context. This paper uses Semi-Global Matching (SGM) with superpixel-based refinement to mitigate the inherent block artefacts and recover lost details. The approach balances the efficiency and accuracy of SGM and adds context-aware segmentation to support more coherent depth inference. The proposed method has been evaluated in three datasets with successful results: In a Mars analogue, the terrain maps obtained show improved structural consistency, particularly in sloped or occlusion-prone regions. Large gaps behind rocks, which are common in raw disparity outputs, are reduced, and surface details like small rocks and edges are captured more accurately. Another two datasets, evaluated to test the method's general robustness and adaptability, show more precise disparity maps and more consistent terrain models, better suited for the demands of autonomous navigation on Mars, and competitive accuracy across both non-occluded and full-image error metrics. This paper outlines the entire terrain modelling process, from finding corresponding features to generating the final 2D navigation maps, offering a complete pipeline suitable for integration in future planetary exploration missions.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.28)
- Asia > Singapore (0.04)
- North America > Canada (0.04)
- Asia > China (0.04)
U-ViLAR: Uncertainty-Aware Visual Localization for Autonomous Driving via Differentiable Association and Registration
Li, Xiaofan, Xu, Zhihao, Wu, Chenming, Yang, Zhao, Zhang, Yumeng, Liu, Jiang-Jiang, Yu, Haibao, Duan, Fan, Ye, Xiaoqing, Wang, Yuan, Li, Shirui, Sun, Xun, Wan, Ji, Wang, Jun
Accurate localization using visual information is a critical yet challenging task, especially in urban environments where nearby buildings and construction sites significantly degrade GNSS (Global Navigation Satellite System) signal quality. This issue underscores the importance of visual localization techniques in scenarios where GNSS signals are unreliable. This paper proposes U-ViLAR, a novel uncertainty-aware visual localization framework designed to address these challenges while enabling adaptive localization using high-definition (HD) maps or navigation maps. Specifically, our method first extracts features from the input visual data and maps them into Bird's-Eye-View (BEV) space to enhance spatial consistency with the map input. Subsequently, we introduce: a) Perceptual Uncertainty-guided Association, which mitigates errors caused by perception uncertainty, and b) Localization Uncertainty-guided Registration, which reduces errors introduced by localization uncertainty. By effectively balancing the coarse-grained large-scale localization capability of association with the fine-grained precise localization capability of registration, our approach achieves robust and accurate localization. Experimental results demonstrate that our method achieves state-of-the-art performance across multiple localization tasks. Furthermore, our model has undergone rigorous testing on large-scale autonomous driving fleets and has demonstrated stable performance in various challenging urban scenarios.
Intention Recognition in Real-Time Interactive Navigation Maps
Zhao, Peijie, Arefin, Zunayed, Meneguzzi, Felipe, Pereira, Ramon Fraga
In this demonstration, we develop IntentRec4Maps, a system to recognise users' intentions in interactive maps for real-world navigation. IntentRec4Maps uses the Google Maps Platform as the real-world interactive map, and a very effective approach for recognising users' intentions in real-time. We showcase the recognition process of IntentRec4Maps using two different Path-Planners and a Large Language Model (LLM). GitHub: https://github.com/PeijieZ/IntentRec4Maps
- South America > Brazil (0.15)
- Europe > United Kingdom > England > Greater Manchester > Manchester (0.05)
- Asia > China > Hong Kong (0.05)
- Europe > United Kingdom > England > Greater London > London > Kensington and Chelsea (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Planning & Scheduling (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.91)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (0.47)
Automatic Navigation Map Generation for Mobile Robots in Urban Environments
Mozzarelli, Luca, Specchia, Simone, Corno, Matteo, Savaresi, Sergio Matteo
A fundamental prerequisite for safe and efficient navigation of mobile robots is the availability of reliable navigation maps upon which trajectories can be planned. With the increasing industrial interest in mobile robotics, especially in urban environments, the process of generating navigation maps has become of particular interest, being a labor intensive step of the deployment process. Automating this step is challenging and becomes even more arduous when the perception capabilities are limited by cost considerations. This paper proposes an algorithm to automatically generate navigation maps using a typical navigation-oriented sensor setup: a single top-mounted 3D LiDAR sensor. The proposed method is designed and validated with the urban environment as the main use case: it is shown to be able to produce accurate maps featuring different terrain types, positive obstacles of different heights as well as negative obstacles. The algorithm is applied to data collected in a typical urban environment with a wheeled inverted pendulum robot, showing its robustness against localization, perception and dynamic uncertainties. The generated map is validated against a human-made map.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.28)
- Europe > Italy > Lombardy > Milan (0.14)
- North America > United States > Illinois > Champaign County > Champaign (0.14)
- (10 more...)
- Research Report (0.64)
- Workflow (0.48)
- Transportation (0.68)
- Automobiles & Trucks (0.46)
Exploring Navigation Maps for Learning-Based Motion Prediction
Schmidt, Julian, Jordan, Julian, Gritschneder, Franz, Monninger, Thomas, Dietmayer, Klaus
The prediction of surrounding agents' motion is a key for safe autonomous driving. In this paper, we explore navigation maps as an alternative to the predominant High Definition (HD) maps for learning-based motion prediction. Navigation maps provide topological and geometrical information on road-level, HD maps additionally have centimeter-accurate lane-level information. As a result, HD maps are costly and time-consuming to obtain, while navigation maps with near-global coverage are freely available. We describe an approach to integrate navigation maps into learning-based motion prediction models. To exploit locally available HD maps during training, we additionally propose a model-agnostic method for knowledge distillation. In experiments on the publicly available Argoverse dataset with navigation maps obtained from OpenStreetMap, our approach shows a significant improvement over not using a map at all. Combined with our method for knowledge distillation, we achieve results that are close to the original HD map-reliant models. Our publicly available navigation map API for Argoverse enables researchers to develop and evaluate their own approaches using navigation maps.
- Asia > Middle East > Jordan (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > California > Santa Clara County > Sunnyvale (0.04)
- Europe > Germany > Baden-Württemberg > Stuttgart Region > Stuttgart (0.04)
- Transportation > Ground > Road (0.51)
- Automobiles & Trucks (0.50)
Dead Reckoning is Still Alive!
Many drivers are highly curious today about the Autonomous Vehicles (AV) dream. Will this dream come true, and when? One of the core technology that needs to be implemented in AV is the inertial navigation system (INS). These systems integrate many sensors together in what we called "sensor fusion" schemes. These sensors include LiDAR, cameras, GPS receivers, Radars, accelerometers, gyroscopes, and many more. The general sensor fusion scheme integrates all the sensors together using a very common algorithm, named the "Kalman Filter", to fuse all sensors optimally (in a mean-squared-error sense).
Scientists develop £2,700 'shoe camera' that detects obstacles
Computer scientists have created an'intelligent' shoe that helps blind and visually-impaired people avoid multiple obstacles. The £2,700 (€3,200) product, called InnoMake, has been developed by Austrian company Tec-Innovation, backed by Graz University of Technology (TU Graz). The product consists of waterproof ultrasonic sensors attached to the tip of each shoe, which vibrate and make noises near obstacles. The closer the wearer gets to an obstacle, the faster the vibration becomes, much like a parking sensor on the back of a vehicle. Tec-Innovation is now working on embedding an AI-powered camera as part of a new iteration of the product.
Ingenious AI significantly improves navigation maps
While Google and other technology giants have their own dynamics to keep the most detailed and up-to-date maps possible, it is an expensive and time-consuming process. And in some areas, the data is limited. To improve this, researchers at MIT and Qatar Computing Research Institute (QCRI) have developed a new machine-learning model based on satellite images that could significantly improve digital maps for GPS navigation. The system, called "RoadTagger," recognizes the types of roads and the number of lanes in satellite images, even in spite of trees or buildings that obscure the view. In the future, the system should recognize even more details, such as bike paths and parking spaces.